**Summary of the Document:** This paper explores the use of Large Language Models (LLMs) in reinforcement learning (RL) for recommender systems, focusing on data efficiency and generative approaches. The authors implement and evaluate RL methods—specifically **Proximal Policy Optimization (PPO)** and **Direct Preference Optimization (DPO)**—using the **WebShop** benchmark, a simulated e-commerce environment with human instructions and product data. ### **Key Findings:** 1. **DPO outperforms PPO** in data efficiency and task performance, achieving a **19% success rate** with just **30 minutes of training** (vs. PPO’s 15% after 2 hours). 2. **Self-learning with generated trajectories** (without human data) matches performance of human-trained agents, suggesting a **low-cost alternative** for RL training. 3. **RL agents optimize long-term user satisfaction**, addressing limitations of traditional supervised learning (e.g., short-term bias). ### **Methodology:** - **WebShop Environment**: Simulates online shopping with 1.18M products and 12K human instructions. - **DPO**: Trains agents using preference-based trajectories without explicit reward models. - **PPO**: Fine-tunes policies with clipped objectives for stability. - **Generative Training**: Uses synthetic trajectories to reduce reliance on costly human data. ### **Implications:** - **Cost Efficiency**: Generative methods enable scalable RL training with minimal human input. - **Recommender Systems**: RL agents can dynamically rank products based on user instructions, improving engagement. **Conclusion**: DPO offers a promising, data-efficient approach for RL in recommenders, with generative training further reducing dependency on human-labeled data. *(Keywords: LLM, Reinforcement Learning, Recommender Systems, DPO, PPO, Generative AI)* **Authors**: Shuang Feng (Stanford), Grace Feng (UCSB) **Published**: KDD’24 Workshop (Accepted July 2024) --- *This summary condenses the paper’s objectives, methods, results, and significance while preserving technical clarity.*